Efficient Private ERM for Smooth Objectives
نویسندگان
چکیده
In this paper, we consider efficient differentially private empirical risk minimization from the viewpoint of optimization algorithms. For strongly convex and smooth objectives, we prove that gradient descent with output perturbation not only achieves nearly optimal utility, but also significantly improves the running time of previous state-of-the-art private optimization algorithms, for both ǫ-DP and (ǫ, δ)-DP. For non-convex but smooth objectives, we propose an RRPSGD (Random Round Private Stochastic Gradient Descent) algorithm, which provably converges to a stationary point with privacy guarantee. Besides the expected utility bounds, we also provide guarantees in high probability form. Experiments demonstrate that our algorithm consistently outperforms existing method in both utility and running time.
منابع مشابه
Differentially Private Empirical Risk Minimization Revisited: Faster and More General
In this paper we study the differentially private Empirical Risk Minimization (ERM) problem in different settings. For smooth (strongly) convex loss function with or without (non)-smooth regularization, we give algorithms that achieve either optimal or near optimal utility bounds with less gradient complexity compared with previous work. For ERM with smooth convex loss function in high-dimensio...
متن کاملMeeting the Challenges of Enterprise Risk Management in Higher Education
enterprise risk management (ERM) in Washington DC with senior officers and trustees from several leading colleges and universities. Approximately 40 leaders—representing large and small as well as public and private institutions—attended the event. (For a list of attendees, see page xx). The summit opened with a presentation on implementing ERM in the for-profit financial services industry. The...
متن کاملEfficient Empirical Risk Minimization with Smooth Loss Functions in Non-interactive Local Differential Privacy
In this paper, we study the Empirical Risk Minimization problem in the non-interactive local model of differential privacy. We first show that if the ERM loss function is (∞, T )-smooth, then we can avoid a dependence of the sample complexity, to achieve error α, on the exponential of the dimensionality p with base 1/α (i.e., α−p), which answers a question in (Smith et al., 2017). Our approach ...
متن کاملAccuracy First: Selecting a Differential Privacy Level for Accuracy Constrained ERM
Traditional approaches to differential privacy assume a fixed privacy requirement ε for a computation, and attempt to maximize the accuracy of the computation subject to the privacy constraint. As differential privacy is increasingly deployed in practical settings, it may often be that there is instead a fixed accuracy requirement for a given computation and the data analyst would like to maxim...
متن کاملPrivate Empirical Risk Minimization Beyond the Worst Case: The Effect of the Constraint Set Geometry
Empirical Risk Minimization (ERM) is a standard technique in machine learning, where a model is selected by minimizing a loss function over constraint set. When the training dataset consists of private information, it is natural to use a differentially private ERM algorithm, and this problem has been the subject of a long line of work [CM08, KST12, JKT12, ST13a, DJW13, JT14, BST14, Ull14]. A pr...
متن کامل